Multi-agent reinforcement learning for Markov routing games: A new modeling paradigm for dynamic traffic assignment
نویسندگان
چکیده
This paper aims to develop a paradigm that models the learning behavior of intelligent agents (including but not limited autonomous vehicles, connected and automated or human-driven vehicles with navigation systems where human drivers follow instructions completely) utility-optimizing goal system's equilibrating processes in routing game among atomic selfish agents. Such can assist policymakers devising optimal operational planning countermeasures under both normal abnormal circumstances. To this end, we Markov (MRG) which each agent learns updates her own en-route path choice policy while interacting others transportation networks. efficiently solve MRG, formulate it as multi-agent reinforcement (MARL) devise mean field deep Q (MF-MA-DQL) approach captures competition The linkage between classical DUE our proposed is discussed. We show shown converge notion predictive dynamic user equilibrium (DUE) when traffic environments are simulated using loading (DNL). In other words, MRG depicts DUEs assuming perfect information deterministic propagated by DNL models. Four examples solved illustrate algorithm efficiency consistency equilibrium, on simple network without spillback, Ortuzar Willumsen (OW) Network, real-world near Columbia University's campus Manhattan New York City.
منابع مشابه
Markov Games as a Framework for Multi-Agent Reinforcement Learning
In the Markov decision process (MDP) formalization of reinforcement learning, a single adaptive agent interacts with an environment defined by a probabilistic transition function. In this solipsistic view, secondary agents can only be part of the environment and are therefore fixed in their behavior. The framework of Markov games allows us to widen this view to include multiple adaptive agents ...
متن کاملMarkov Games of Incomplete Information for Multi-Agent Reinforcement Learning
Partially observable stochastic games (POSGs) are an attractive model for many multi-agent domains, but are computationally extremely difficult to solve. We present a new model, Markov games of incomplete information (MGII) which imposes a mild restriction on POSGs while overcoming their primary computational bottleneck. Finally we show how to convert a MGII into a continuous but bounded fully ...
متن کاملA Distributed Reinforcement Learning Scheme for Network Markov Games as a Framework for Multi-agent Reinforcement Learning. 8.2 Discussion
Consider an electronic market where agents can interact and trade. The agents involved in the market are completely autonomous and act on behalf of their masters. In such a multi-agent system, where other agents may be potential partners, or competing opponents, an agent should have the ability to identify the other agents' intentions and goals and to be able to predict the others' future behav...
متن کاملReinforcement Learning in Multi-agent Games
This article investigates the performance of independent reinforcement learners in multiagent games. Convergence to Nash equilibria and parameter settings for desired learning behavior are discussed for Q-learning, Frequency Maximum Q value (FMQ) learning and lenient Q-learning. FMQ and lenient Q-learning are shown to outperform regular Q-learning significantly in the context of coordination ga...
متن کاملImproved Multi-Agent Reinforcement Learning for Minimizing Traffic Waiting Time
This paper depict using multi-agent reinforcement learning (MARL) algorithm for learning traffic pattern to minimize the traveling time or maximizing safety and optimizing traffic pattern (OTP). This model provides a description and solution to optimize traffic pattern that use multi-agent based reinforcement learning algorithms. MARL uses multi agent structure where vehicles and traffic signal...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Transportation Research Part C-emerging Technologies
سال: 2022
ISSN: ['1879-2359', '0968-090X']
DOI: https://doi.org/10.1016/j.trc.2022.103560